Alignment Research, Model Robustness, Adversarial Examples, Risk Assessment

Cross-Domain Misalignment Generalization: Contextual Role Inference
lesswrong.com·18h
🎯Embedding Models
15 Hot Takes That Cut Through the AI Hype
creatoreconomy.so·14h
🛡️AI Security
[ARC-AGI-2 SoTA] Efficient Evolutionary Program Synthesis
ctpang.substack.com·23h·
Discuss: Substack
🏆LLM Benchmarking
Today we’re releasing research with @apolloaievals.
threadreaderapp.com·12h
🛡️AI Security
Bring us your LLMs: why peer review is good for AI models
nature.com·13h
🏆LLM Benchmarking
Is It Real, or Is It AI?
cacm.acm.org·14h
💳Content Monetization
The rise of AI-ready private clouds
infoworld.com·20h
🖥GPUs
A new study finds AI tools are often unreliable, overconfident and one-sided
techxplore.com·12h
🔍AI Interpretability
Tau² Benchmark: How a Prompt Rewrite Boosted GPT-5-Mini by 22%
quesma.com·16h·
Discuss: Hacker News
🏆LLM Benchmarking
Don’t Trust the Scroll: Stop AI Agents from Running Code
pub.towardsai.net·17h
🛡️AI Security
Delegation to AI can increase dishonest behavior
phys.org·14h
🛡️AI Security
AI Revives Chipmaking as Tech’s Core Engine
semiwiki.com·14h
🔬Chip Fabrication
Why Culture Is the First Line of Defense in the Age of Agentic AI
nordot.app·18h
🛡️AI Security
AI model forecasts disease risk decades in advance
medicalxpress.com·14h
🛡️AI Security
DeepSeek-R1 on Nature: How Pure Reinforcement Learning Unlocks LLM Reasoning
reddit.com·5h·
Discuss: r/LocalLLaMA
🏗️LLM Infrastructure
3. Americans on the risks, benefits of AI – in their own words
pewresearch.org·15h
🛡️AI Security
Partnering with Irregular: Ahead of the Curve
sequoiacap.com·16h
🛡️AI Security
The AI doomers are not making an argument. They’re selling a worldview.
vox.com·17h
🛡️AI Security
Visual Exploration of Gradient Descent (many images)
lesswrong.com·16h
🎯Qdrant
The Center for AI Policy Has Shut Down
lesswrong.com·18h
📜Semiconductor Policy